# Video Generation

ASMR.so
Asmr.so
ASMR.so is a platform based on advanced VEO3 AI technology that allows users to quickly generate professional ASMR videos. The product supports multiple ASMR types, including whispers, tapping, and natural sounds, aiming to provide users with a relaxing and enjoyable experience. Its main advantages include fast video generation (usually completed within 2 minutes), high-definition quality, and a user-friendly operation process. It is suitable for video creators, ASMR enthusiasts, and users who need relaxing content. The platform also offers a flexible credit system, allowing users to choose packages according to their needs. In terms of pricing, there are free trials and paid packages available.
Video Generation
37.0K
AI ASMR
AI ASMR
AI ASMR Generator is a tool that uses AI technology to generate ASMR videos. It helps users quickly create high-quality ASMR videos, providing a richer experience and stimulation.
AI
37.0K
UnificAlly
Unifically
UnificAlly is an AI API service platform that offers innovative AI models and API services at competitive prices. Users can access the platform and select various advanced AI models, such as GPT 4.1, Suno, Higgsfield, etc., for video generation, image creation, and music composition. UnificAlly is committed to providing cost-effective AI services and is known for its fast and reliable API responses, easy-to-integrate REST APIs, and comprehensive documentation and examples.
API
37.0K
A2E Free and Uncensored AI Videos
A2E Free And Uncensored AI Videos
a2e.ai is an AI tool that offers functions such as AI Avatar, lip sync, voice cloning, and text-to-video. The product has advantages such as high clarity, high consistency, and efficient generation speed, and is suitable for various scenarios, providing a complete set of AI Avatar tools.
Lip Sync
37.0K
Vidduo
Vidduo
The AI video generator uses leading industry image-to-video AI technology, intelligently selects the best models, generates 1080p videos, supports multi-shot shooting, has diverse styles, and produces smooth motion. Key advantages include fast generation of high-quality videos, support for complex scenes and lens movement control, and suitability for designers, content creators, and other users.
Video Generation
37.8K
Seedance AI
Seedance AI
Seedance AI is a video generator by ByteDance that uses Seedance 1.0 Pro technology to achieve professional cinematic quality. Users can generate cinematic-quality videos with simple text or image prompts.
Video Generation
38.1K
Chinese Picks
HunyuanCustom
Hunyuancustom
HunyuanCustom is a multimodal customized video generation framework designed to generate specific-topic videos based on user-defined conditions. The technology excels in identity consistency and support for multiple input modes, capable of processing text, images, audio, and video inputs, applicable to various scenarios such as virtual human advertising and video editing.
Multimodal
38.9K
PixVerse-MCP
Pixverse MCP
PixVerse-MCP is a tool that allows users to access PixVerse's latest video generation models through applications that support the Model Context Protocol (MCP). This product offers features such as text-to-video generation and is suitable for creators and developers to generate high-quality videos anywhere. The PixVerse platform requires API credits, which users need to purchase themselves.
Video Production
39.7K
English Picks
AvatarFX
Avatarfx
AvatarFX is a cutting-edge AI platform focused on interactive storytelling. Users can quickly generate vivid, realistic character videos by uploading images and selecting sounds. Its core technology is based on a DiT diffusion video generation model, enabling efficient generation of high-fidelity, temporally consistent videos, especially suitable for creations requiring multiple characters and dialogue scenes. The product is positioned to provide creators with tools to help them realize the limitless possibilities of their imagination.
Video Generation
39.2K
Fresh Picks
Vidu Q1
Vidu Q1
Vidu Q1, launched by Shengshu Technology, is a domestically produced video generation large language model designed for video creators. It supports high-definition 1080p video generation and features cinematic camera movements and start/end frame functionality. This product ranked first in the VBench-1.0 and VBench-2.0 evaluations, offering exceptional value for money at only one-tenth the price of competitors. It is suitable for film, advertising, animation, and other fields, significantly reducing production costs and improving creative efficiency.
Video Production
37.0K
Fresh Picks
SkyReels-V2
Skyreels V2
SkyReels-V2 is the world's first infinite-length movie generation model using a diffusion forcing framework, released by Kunlun Wanwei's SkyReels team. The model achieves synergistic optimization by combining multimodal large language models, multi-stage pre-training, reinforcement learning, and the diffusion forcing framework, overcoming significant challenges in traditional video generation technology regarding prompt following, visual quality, motion dynamics, and video duration coordination. It not only provides content creators with powerful tools but also unlocks unlimited possibilities for video storytelling and creative expression using AI.
Video Production
39.5K
Chinese Picks
Wan2.1-FLF2V-14B
Wan2.1 FLF2V 14B
Wan2.1-FLF2V-14B is an open-source, large-scale video generation model designed to advance the field of video generation. This model excels in multiple benchmark tests, supports consumer-grade GPUs, and efficiently generates 480P and 720P videos. It performs exceptionally well in various tasks, including text-to-video and image-to-video, possessing strong visual-text generation capabilities suitable for diverse real-world applications.
Video Production
39.2K
FramePack
Framepack
FramePack is an innovative video generation model designed to improve the quality and efficiency of video generation by compressing the context of input frames. Its main advantage lies in addressing the drift problem in video generation, maintaining video quality through a bidirectional sampling method, and being suitable for users who need to generate long videos. This technology is based on in-depth research and experiments on existing models to improve the stability and coherence of video generation.
Video Production
37.8K
Pusa
Pusa
Pusa introduces an innovative approach to video diffusion modeling through frame-level noise control, enabling high-quality video generation suitable for various tasks (text-to-video, image-to-video, etc.). With its superior motion fidelity and efficient training process, the model offers an open-source solution for convenient video generation.
Video Production
38.4K
SkyReels-A2
Skyreels A2
SkyReels-A2 is a framework based on a video diffusion transformer that allows users to synthesize and generate video content. The model, leveraging deep learning technology, provides flexible creative capabilities suitable for various video generation applications, especially in animation and special effects production. The product's advantages lie in its open-source nature and efficient model performance, making it suitable for researchers and developers. It is currently free of charge.
Video Production
38.6K
Chinese Picks
OmniTalker
Omnitalker
OmniTalker is a unified framework proposed by Alibaba's Tongyi Lab with the aim of generating audio and video in real time to enhance human-computer interaction experiences. Its innovation lies in solving common issues in traditional text-to-speech and speech-driven video generation methods, such as out-of-sync audio-video, inconsistent styles, and system complexity. OmniTalker adopts a dual-branch diffusion transformer architecture, achieving high-fidelity audio-video outputs while maintaining efficiency. Its real-time inference speed reaches 25 frames per second, making it suitable for various interactive video chat applications and enhancing user experiences.
Video Generation
38.6K
DreamActor-M1
Dreamactor M1
DreamActor-M1 is a human animation framework based on Diffusion Transformer (DiT), designed to achieve fine-grained overall controllability, multi-scale adaptability, and long-term temporal consistency. The model, through blending guidance, can generate high-expressiveness and realistic human videos suitable for various scenarios from portrait to full-body animation. Its main advantages lie in its high fidelity and identity preservation, bringing new possibilities to human behavior animation.
Video Production
37.5K
GAIA-2
GAIA 2
GAIA-2 is an advanced video generation model developed by Wayve, designed to provide diverse and complex driving scenarios for autonomous driving systems to improve safety and reliability. The model addresses the limitations of relying on real-world data collection by generating synthetic data, capable of creating various driving situations, including both regular and edge cases. GAIA-2 supports the simulation of various geographical and environmental conditions, helping developers quickly test and verify autonomous driving algorithms without high costs.
Video Production
39.2K
AccVideo
Accvideo
AccVideo is a novel and efficient distillation method that accelerates the inference speed of video diffusion models through synthetic datasets. This model can achieve an 8.5-fold speed improvement in video generation while maintaining similar performance. It uses a pre-trained video diffusion model to generate multiple effective denoising trajectories, thereby optimizing data usage and the generation process. AccVideo is especially suitable for scenarios requiring efficient video generation, such as film production and game development, and is suitable for researchers and developers.
Video Production
63.5K
Video-T1
Video T1
Video-T1 is a video generation model that significantly improves the quality and consistency of generated videos through test-time scaling (TTS) technology. This technology allows more computational resources to be used during inference, thereby optimizing the generation results. Compared to traditional video generation methods, TTS can provide higher generation quality and richer content expression, suitable for the digital creation field. This product is mainly aimed at researchers and developers, and the price information is not clear.
Video Production
54.4K
vivago.ai
Vivago.ai
vivago.ai is a free AI generation tool and community that provides text-to-image, image-to-video, and other functions, making creation simpler and more efficient. Users can generate high-quality images and videos for free, with support for various AI editing tools to facilitate creation and sharing. The platform aims to provide creators with easy-to-use AI tools to meet their visual creation needs.
Image Generation
64.9K
Long Context Tuning (LCT)
Long Context Tuning (LCT)
Long Context Tuning (LCT) aims to bridge the gap between current single-generation capabilities and real-world narrative video production. This technology directly learns scene-level consistency through data-driven methods, supporting interactive multi-shot development and synthetic generation, applicable to all aspects of video production.
Video Production
70.1K
MM_StoryAgent
MM StoryAgent
MM_StoryAgent is a story video generation framework based on the multi-agent paradigm. It combines multiple modalities such as text, images, and audio to generate high-quality story videos through a multi-stage process. The core advantage of this framework lies in its customizability; users can customize expert tools to improve the generation quality of each component. Furthermore, it provides a list of story themes and evaluation criteria to facilitate further story creation and evaluation. MM_StoryAgent is primarily aimed at creators and businesses that need to efficiently generate story videos; its open-source nature allows users to extend and optimize it according to their own needs.
Video Production
69.8K
Flat Color - Style
Flat Color Style
Flat Color - Style is a LoRA model specifically designed for generating flat-color style images and videos. Trained on the Wan Video model, it features a unique lineless, low-depth effect, suitable for anime, illustration, and video generation. The model's main advantages are its ability to reduce color bleeding, enhance black expression, and provide high-quality visual effects. It's suitable for scenarios requiring simple, flat designs, such as anime character design, illustration creation, and video production. This model is offered free to users to help creators quickly achieve modern and simple visual works.
Image Generation
71.5K
HunyuanVideo-I2V
Hunyuanvideo I2V
HunyuanVideo-I2V is an open-source image-to-video generation model developed by Tencent based on the HunyuanVideo architecture. This model effectively integrates reference image information into the video generation process through image latent splicing technology, supports high-resolution video generation, and provides customizable LoRA effect training functions. This technology is of great significance in the field of video creation, helping creators quickly generate high-quality video content and improve creation efficiency.
Video Production
92.5K
Wan2GP
Wan2gp
Wan2GP is an improved version based on Wan2.1, aiming to provide an efficient and low-memory video generation solution for low-configuration GPU users. The model, through optimized memory management and accelerated algorithms, enables ordinary users to quickly generate high-quality video content on consumer-grade GPUs. It supports multiple tasks, including text-to-video, image-to-video, and video editing, and features a powerful video VAE architecture capable of efficiently handling 1080P videos. The emergence of Wan2GP lowers the barrier to entry for video generation technology, allowing more users to easily learn and apply it to real-world scenarios.
Video Production
73.7K
hunyuan-video-keyframe-control-lora
Hunyuan Video Keyframe Control Lora
HunyuanVideo Keyframe Control LoRA is an adapter for the HunyuanVideo T2V model, focusing on keyframe video generation. It modifies the input embedding layer to effectively integrate keyframe information and applies Low-Rank Adaptation (LoRA) technology to optimize linear and convolutional input layers, enabling efficient fine-tuning. This model allows users to precisely control the starting and ending frames of the generated video by defining keyframes, ensuring seamless integration with the specified keyframes and enhancing video coherence and narrative. It has significant application value in video generation, particularly excelling in scenarios requiring precise control over video content.
Video Production
61.5K
TheoremExplainAgent
Theoremexplainagent
TheoremExplainAgent is an AI-powered model focused on generating detailed multimodal explanatory videos for mathematical and scientific theorems. It helps users gain a deeper understanding of complex concepts by combining text and visual animations. This product uses Manim animation technology to generate long videos exceeding 5 minutes, addressing the shortcomings of traditional text explanations, particularly excelling in revealing reasoning errors. Primarily aimed at the education sector to enhance learners' understanding of STEM theorems, its pricing and commercialization strategy are not yet clearly defined.
Education
62.4K
ComfyUI-WanVideoWrapper
Comfyui WanVideoWrapper
ComfyUI-WanVideoWrapper provides ComfyUI nodes for WanVideo, enabling users to leverage WanVideo's functionality within the ComfyUI environment for video generation and processing. Developed in Python, this tool supports efficient content creation and video generation, suitable for users needing to quickly produce video content.
Video Production
101.0K
Wan2.1
Wan2.1
Wan2.1 is an open-source, advanced, large-scale video generation model designed to push the boundaries of video generation technology. Through innovative spatio-temporal variational autoencoders (VAEs), scalable training strategies, large-scale data construction, and automated evaluation metrics, it significantly improves model performance and versatility. Wan2.1 supports multiple tasks, including text-to-video, image-to-video, and video editing, and can generate high-quality video content. The model has demonstrated superior performance in several benchmark tests, even surpassing some closed-source models. Its open-source nature allows researchers and developers to freely use and extend the model, making it suitable for various applications.
Video Production
67.6K
Featured AI Tools
Flow AI
Flow AI
Flow is an AI-driven movie-making tool designed for creators, utilizing Google DeepMind's advanced models to allow users to easily create excellent movie clips, scenes, and stories. The tool provides a seamless creative experience, supporting user-defined assets or generating content within Flow. In terms of pricing, the Google AI Pro and Google AI Ultra plans offer different functionalities suitable for various user needs.
Video Production
42.0K
NoCode
Nocode
NoCode is a platform that requires no programming experience, allowing users to quickly generate applications by describing their ideas in natural language, aiming to lower development barriers so more people can realize their ideas. The platform provides real-time previews and one-click deployment features, making it very suitable for non-technical users to turn their ideas into reality.
Development Platform
44.2K
ListenHub
Listenhub
ListenHub is a lightweight AI podcast generation tool that supports both Chinese and English. Based on cutting-edge AI technology, it can quickly generate podcast content of interest to users. Its main advantages include natural dialogue and ultra-realistic voice effects, allowing users to enjoy high-quality auditory experiences anytime and anywhere. ListenHub not only improves the speed of content generation but also offers compatibility with mobile devices, making it convenient for users to use in different settings. The product is positioned as an efficient information acquisition tool, suitable for the needs of a wide range of listeners.
AI
41.7K
MiniMax Agent
Minimax Agent
MiniMax Agent is an intelligent AI companion that adopts the latest multimodal technology. The MCP multi-agent collaboration enables AI teams to efficiently solve complex problems. It provides features such as instant answers, visual analysis, and voice interaction, which can increase productivity by 10 times.
Multimodal technology
42.8K
Chinese Picks
Tencent Hunyuan Image 2.0
Tencent Hunyuan Image 2.0
Tencent Hunyuan Image 2.0 is Tencent's latest released AI image generation model, significantly improving generation speed and image quality. With a super-high compression ratio codec and new diffusion architecture, image generation speed can reach milliseconds, avoiding the waiting time of traditional generation. At the same time, the model improves the realism and detail representation of images through the combination of reinforcement learning algorithms and human aesthetic knowledge, suitable for professional users such as designers and creators.
Image Generation
41.4K
OpenMemory MCP
Openmemory MCP
OpenMemory is an open-source personal memory layer that provides private, portable memory management for large language models (LLMs). It ensures users have full control over their data, maintaining its security when building AI applications. This project supports Docker, Python, and Node.js, making it suitable for developers seeking personalized AI experiences. OpenMemory is particularly suited for users who wish to use AI without revealing personal information.
open source
42.0K
FastVLM
Fastvlm
FastVLM is an efficient visual encoding model designed specifically for visual language models. It uses the innovative FastViTHD hybrid visual encoder to reduce the time required for encoding high-resolution images and the number of output tokens, resulting in excellent performance in both speed and accuracy. FastVLM is primarily positioned to provide developers with powerful visual language processing capabilities, applicable to various scenarios, particularly performing excellently on mobile devices that require rapid response.
Image Processing
40.8K
Chinese Picks
LiblibAI
Liblibai
LiblibAI is a leading Chinese AI creative platform offering powerful AI creative tools to help creators bring their imagination to life. The platform provides a vast library of free AI creative models, allowing users to search and utilize these models for image, text, and audio creations. Users can also train their own AI models on the platform. Focused on the diverse needs of creators, LiblibAI is committed to creating inclusive conditions and serving the creative industry, ensuring that everyone can enjoy the joy of creation.
AI Model
6.9M
AIbase
Empowering the Future, Your AI Solution Knowledge Base
© 2025AIbase